What is Decidable about Partially Observable Markov Decision Processes with ω-Regular Objectives

نویسندگان

  • Krishnendu Chatterjee
  • Martin Chmelik
  • Mathieu Tracol
چکیده

We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specified as parity objectives. The class of ω-regular languages extends regular languages to infinite strings and provides a robust specification language to express all properties used in verification, and parity objectives are canonical forms to express ω-regular conditions. The qualitative analysis problem given a POMDP and a parity objective asks whether there is a strategy to ensure that the objective is satisfied with probability 1 (resp. positive probability). While the qualitative analysis problems are known to be undecidable even for very special cases of parity objectives, we establish decidability (with optimal complexity) of the qualitative analysis problems for POMDPs with all parity objectives under finitememory strategies. We establish asymptotically optimal (exponential) memory bounds and EXPTIMEcompleteness of the qualitative analysis problems under finite-memory strategies for POMDPs with parity objectives.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

What is Decidable about Partially Observable Markov Decision Processes with omega-Regular Objectives

We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specified as parity objectives. The class of ω-regular languages extends regular languages to infinite strings and provides a robust specification language to express all properties used in verification, and parity objectives are canonical forms to express ω-regular conditions. The qualitative analysis...

متن کامل

Multiple-Environment Markov Decision Processes

We introduce Multi-Environment Markov Decision Processes (MEMDPs) which are MDPs with a set of probabilistic transition functions. The goal in a MEMDP is to synthesize a single controller with guaranteed performances against all environments even though the environment is unknown a priori. While MEMDPs can be seen as a special class of partially observable MDPs, we show that several verificatio...

متن کامل

Deciding the Value 1 Problem for ]-acyclic Partially Observable Markov Decision Processes

The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there strategies that achieve the reachability objective with probability arbitrarily close to 1? This problem was shown undecidable recently. Our contribution is to introduce a class of partially ob...

متن کامل

Deciding the Value 1 Problem for $\sharp$ -acyclic Partially Observable Markov Decision Processes

The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there observational strategies that achieve the reachability objective with probability arbitrarily close to 1? This problem was shown undecidable recently. Our contribution is to introduce a class o...

متن کامل

A Specification Logic for Reasoning about Actions and Explicit Observations

We propose a logic that can express explicit observation constants, as a bridge towards reasoning about action using partially observable Markov decision processes (POMDPs). We introduce a new modal operator [o | α]—informally, [o | α]φ means ‘after perceiving observation o, given α was performed, necessarily φ’. In this paper, we focus on the challenges concerning sensing with explicit observa...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • J. Comput. Syst. Sci.

دوره 82  شماره 

صفحات  -

تاریخ انتشار 2016